The 10th international conference on Machine Learning and Artificial Intelligence applications,
taking place in person in Prague and online.

Machine Learning Prague 2025

, 2025

Registration

World class expertise and practical content packed in 3 days!

You can look forward to an excellent lineup of 45 international experts in ML and AI business and academic applications at ML Prague 2025. They will present advanced practical talks, hands-on workshops, and other forms of interactive content to you.

What to expect

  • 1000+ Attendees
  • 3 Days
  • 45 Speakers
  • 10 Workshops

Phenomenal Confirmed speakers

Iryna Gurevych

Full professor, Technical University of Darmstadt

Iryna Gurevych is a Full Professor (W3) at the Computer Science Department of the Technical University Darmstadt, Germany and head of the UKP Lab. She has a strong background in information extraction, semantic text processing, machine learning and innovative applications of NLP to social sciences and humanities.

Since 2014, she is co-director of the Centre for the Digital Foundation of Research in the Humanities, Social, and Educational Sciences (CEDIFOR[3]), which is funded by the Federal Ministry of Education and Research.[4] The following year, she founded the research training group AIPHES[5] (Adaptive Information Preparation from Heterogeneous Sources) funded by the German Research Foundation. Since 2020, Gurevych is the director of CA-SG,[6] a research initiative "Content Analytics for the Social Good" of the Rhine-Main Universities and co-director of the Natural Language Processing (NLP) program of ELLIS, a European Network of Excellence in Machine Learning.

In 2020, Gurevych was awarded as a Fellow of the international scientific Association for Computational Linguistics (ACL) for her outstanding contributions in the field of Natural Language Processing and Machine Learning.[7] On January 1, 2021, Gurevych has taken over the office of Vice-president-elect and becomes president of the most important international organization in computational linguistics in 2023: the Association for Computational Linguistics (ACL).[8]

Gurevych receives the first LOEWE-professorship of the LOEWE programme, a Hessian research funding programme in Germany, in March 2021.

Gurevych's research interests include Natural Language Processing, Machine Learning, Multimodal Data Analysis, Digital Humanities, and Computational Social Science.

Ivan Cimrák

Lead Researcher, University of Zilina

Ivan Cimrák is a researcher and university lecturer at the University of Žilina in Slovakia, specializing in applied mathematics and informatics with a focus on modeling the separation of circulating cancer cells and the application of artificial intelligence in biomedicine. 

He has held postdoctoral positions at Ghent University in Belgium and St. Pölten University of Applied Sciences in Austria. 

Dr. Cimrák has been the recipient of an individual Marie Curie EU grant and has successfully secured several national grants, underscoring his contributions to the scientific community. 

In 2018, he co-authored the book "Computational Blood Cell Mechanics: Road Towards Models and Biomedical Applications," which presents a comprehensive study on modeling blood cells and their behavior under flow conditions. 

His research includes work on classifying red blood cells using time-distributed convolutional neural networks from simulated videos, as well as developing curated datasets for red blood cell tracking in microfluidic devices. 

Dr. Cimrák leads a research lab at the University of Žilina, where he mentors a team of researchers and students in advancing computational methods in biomedicine. 

He has been recognized as an exceptional figure in Slovak science, being nominated to be between four finalists for the ESET Science Award in the category of Outstanding Personality of University Education. 

The award Science and Technology Award 2024 in the category Personality of Science and Technology is a prestigious recognition of Dr. Ivan Cimrák's significant contributions to the field of applied mathematics, biomedicine, and artificial intelligence. This honor awarded by Ministry of Education, Research, Development and Youth of the Slovak Republic highlights his dedication to advancing scientific understanding and innovation, solidifying his status as a leading figure in Slovak science and technology. 

Dr. Cimrák actively participates in conferences and discussions related to technology in oncology, contributing to the advancement of biomedical applications of artificial intelligence. 

His work continues to bridge the gap between computational modeling and practical biomedical applications, enhancing the understanding and treatment of complex medical conditions.

Practical & Inspiring Program

Friday
Workshops

O2 Universum, Českomoravská 2345/17a, 190 00, Praha (workshops won't be streamed)

Registration

Room D2 Room D3 Room D4 Room D6 Room D7
coffee break

Utilizing Large Language Models for improved anti-tracking in web browsers

Room D2

Humera Noor Minhas, Digital Munich

Online tracking remains a significant privacy concern for internet users. Current solutions while effective have limitations in terms of coverage maintenance and precision. This workshop aims to leverage the power of LLMs to create a more robust adaptive and efficient anti-tracking system. We will explore the architecture of an LLM-based anti-tracking system developing the data pipeline and exploring how these models can be fine-tuned to analyze network requests page content and user interactions in real-time. The system's ability to understand the semantic context of web elements allows for more accurate identification of tracking attempts reducing false positives while improving detection rates of sophisticated trackers. A key focus will be on the practical challenges of implementing such a system within the constraints of a web browser environment. We'll discuss strategies for optimizing LLM inference to meet the real-time demands of browsing balancing accuracy with performance.

Accelerating AI Through Human Knowledge: Teaching to Imitate Experts and Win on the Race Track

Room D3

Alexander Buchelt, St. Pölten University of Applied Sciences
Sebastian Eresheim, St. Pölten University of Applied Sciences
Tobias Kietreiber, St. Pölten University of Applied Sciences

In this 3-hour hands-on workshop participants will explore the exciting world of Imitation Learning a powerful technique in artificial intelligence that allows agents to mimic expert behavior and excel in complex environments. Building on the fundamentals of Reinforcement Learning this workshop introduces the theory behind Imitation Learning and demonstrates how it can be applied to solve real-world problems efficiently. By guiding AI through expert demonstration imitation learning accelerates training especially in environments where traditional reinforcement learning might be time-consuming or difficult. Imitation Learning is crucial for AI systems that need to learn from limited data or human expertise such as autonomous driving robotics and gaming. In contrast to trial-and-error methods in reinforcement learning imitation learning allows models to replicate the strategies of experienced individuals drastically reducing training time and improving performance. Attendees will gain a deep understanding of how this approach combines the best of both supervised and reinforcement learning creating smarter faster decision-making systems.

Introduction to Algorithmic Trading: Hands-On Strategy Implementation with Real-World Data

Room D4

Szymon Bubak, Jiai
Jérémy Cochoy, Redstone Solution OÜ

Algorithmic trading has become a cornerstone of financial markets with automation and data-driven strategies driving the majority of transactions. This workshop provides a comprehensive hands-on introduction to the world of algorithmic trading aimed at students and professionals interested in financial markets who want to move beyond academic exercises and engage with real-world scenarios. Participants will develop a trading strategy using real-world data such as Bitcoin prices or stocks from the S&P 500 with the initial focus on achieving profitability under the assumption of no transaction fees. The 3-hour session will guide participants through the end-to-end process of designing implementing and backtesting a deep learning-based trading strategy. The workshop will be structured as follows: - Data Preparation: Participants will learn how to source and preprocess financial data preparing it for model input. - Feature Extraction: We will introduce simple features from the data to feed into a deep learning model. - Loss Function Design: The workshop will cover how to design a loss function tailored to trading strategies and objectives. - Model Training: Participants will implement and train a deep learning model using PyTorch. The workshop will focus on how to properly train models in low-data environments ensuring the model generalizes effectively by employing robust train/validation/test splits. Backtesting: We will backtest the model’s performance allowing participants to understand its strengths and limitations in various market conditions. Rather than dividing the session into theoretical and practical segments the entire workshop will seamlessly integrate theory and implementation. Participants will work on a live trading scenario continuously applying new knowledge as they progress through the workshop. By the end they will have created a complete training pipeline—from data preparation to model training and backtesting. A critical aspect of this workshop will be understanding the practical challenges of implementing algorithmic trading strategies in real-world markets. We will discuss the limitations of deep learning models when applied to financial data including how to mitigate overfitting in low-data environments. Additionally participants will explore the impact of transaction fees slippage and other market inefficiencies learning how these factors affect profitability. The workshop is designed for participants with a basic understanding of Python and machine learning but no prior experience with algorithmic trading is required. By the end of the session attendees will have gained practical skills in designing and implementing a trading strategy along with valuable insights into the intricacies of algorithmic trading in financial markets. Participants will be provided with all necessary code templates and data and they are expected to bring their own laptops to engage fully in the hands-on aspects of the workshop. This workshop offers a unique opportunity to bridge the gap between theoretical learning and real-world financial applications equipping participants with the tools and knowledge to pursue further exploration in algorithmic trading.

InstructLab: plug your knowledge into a model easily

Room D6

Štefan Bunčiak, Red Hat
Tomáš Tomeček, Red Hat

During this hands-on exercise you will learn what is InstructLab and how you can leverage it to easily extend Large Language Models with your data and run them on your infrastructure. The tool makes it easy to download run and chat with models locally on your laptop. InstructLab is a fully open-source project from Red Hat and the MIT-IBM Watson AI Lab that introduces Large-scale Alignment for chatBots (LAB). The paper behind it: https://arxiv.org/abs/2403.01081 The LAB method is driven by taxonomies which are largely created manually and with care. For a taxonomy you supply InstructLab then can generate synthetic data used to train a model. Everyone who has experience with LLMs can greatly benefit from this workshop. We will create our own knowledge documents use InstructLab to generate synthetic data out of them train a model from the data and chat with them.

Beyond Real-World Limitations - Mastering Synthetic Data Generation for Enhanced ML Performance

Room D7

Tomáš Sikora, Emplifi
Kryštof Šaml, Emplifi

As machine learning pushes into new frontiers the demand for diverse and representative datasets frequently outpaces the availability of real-world data. This workshop explores a spectrum of advanced techniques for synthetic data generation from traditional methods to cutting-edge AI agent-driven approaches. By embracing the principles of data-centric AI we'll start with traditional data generation techniques and progress to innovative AI agent-driven strategies. Throughout the workshop we'll demonstrate techniques that promise to overcome limitations in real-world datasets reshape the data landscape in ML and rigorously evaluate the quality of the generated data. Workshop Overview: - Setting the Stage: The Data Challenge in ML We'll begin with a brief historical context touching on pre-LLM approaches to synthetic data generation. This background will highlight the impact of recent advancements and set the stage for our deep dive into modern techniques. - Data-Centric AI: A Paradigm Shift We'll explore how the data-centric AI movement is reshaping our approach to machine learning. Participants will learn why focusing on data quality and targeted synthetic data generation can often yield better results than simply increasing dataset size or model complexity. -Evolution of Synthetic Data Generation with AI -LLM-Powered Data Creation: Leveraging large language models for synthetic data generation. -Multi-Agent Systems: Advancing to manually designed agent ecosystems for nuanced and accurate data production. -Automated Agent Workflows: Exploring the cutting edge with self-optimizing agent interactions for superior data quality. -Robust Evaluation in the Data-Centric Paradigm We'll emphasize rigorous evaluation techniques aligned with data-centric AI principles ensuring the effectiveness of synthetic data in real-world ML applications. This workshop is ideal for ML practitioners researchers and data scientists looking to overcome data quality and scarcity challenges. Participants should have a basic understanding of machine learning concepts and some experience with programming in Python. By the end of this workshop attendees will have a comprehensive understanding of how data-centric AI principles can be applied to synthetic data generation from LLM-based techniques to state-of-the-art agent-driven approaches. They'll be equipped with practical skills to implement these strategies potentially revolutionizing how they tackle data-related challenges in ML projects.

Lunch
coffee break

3D reconstruction from Images and their application

Room D2

Varun Burde, CIIRC, Czech Technical University
Artem Moroz, CIIRC, Czech Technical University
Vit Zeman, CIIRC, Czech Technical University

This workshop will delve into recent advances in 3D computer vision and provide participants with practical hands-on experience in generating 3D reconstructions from image data. Attendees will explore cutting-edge techniques including neural radiance fields (NeRFs) Gaussian splatting multi-view stereo (MVS) and structure from motion (SfM) for surface reconstruction. The session will cover the fundamentals of 3D reconstruction focusing on how modern algorithms transform 2D images into detailed and accurate 3D models. Additionally participants will learn the essential steps of dataset creation and optimization for training advanced 3D reconstruction methods. A key feature of the workshop will involve capturing a set of images of objects and demonstrating how to systematically collect and organize data to ensure high-quality 3D model generation. Attendees will gain experience in building datasets tailored to different 3D reconstruction techniques such as NeRF and Gaussian splatting and optimizing them for improved accuracy and visual fidelity. This workshop is ideal for researchers engineers and enthusiasts seeking to understand the latest in 3D vision technologies with applications ranging from augmented reality and robotics to digital content creation.

Parallel Genetic Algorithms in Python

Room D3

Jakub Tomasz Gnyp, International Centre for Theory of Quantum Technologies
Agata Gurzynska, PricewaterhouseCoopers

In this workshop we delve into the construction and implementation of parallel genetic algorithms (PGAs) using Python. Genetic algorithms (GAs) and evolutionary algorithms in general are powerful tools for solving optimization problems and when parallelized they offer significant speedups and efficiency improvements. Participants who learn PGAs will also be able to apply them in reinforcement learning. The workshop will have a limited amount of mathematics - instead the focus will be on both the idea behind PGAs and practical coding skills. Starting with the PyGAD library its uses and limitations will be discussed and presented with easy-to-understand examples. Later key aspects of parallel programming will be introduced such as recognizing CPU- and I/O-bound operations and the use of processes and threads respectively. Global lock in Python will be addressed as well as racing conditions. Therefore a basic understanding and implementation of locks barriers flags and shared memory in general will be achieved. To illustrate the practical applications of parallel genetic algorithms apart from minor examples the workshop features three major case studies. The first involves solving a labyrinth demonstrating how a parallel genetic algorithm can efficiently navigate complex search spaces and de facto interact with an environment. Participants will observe how the parallelization of GAs can lead to faster convergence on optimal paths compared to sequential approaches. Diversity in population will be addressed as well. The second case study explores the application of parallel genetic algorithms in quantum cryptography. In this domain GAs can optimize parameters for quantum key distribution protocols enhancing especially efficiency. By parallelizing the algorithm we can tackle the computational challenges of the vast solution spaces inherent in quantum cryptographic systems. The BB84 protocol will be the protocol in question explained without the quantum mechanics' mathematical rigor and the essentials of the protocol will already be implemented. The third and last case study will be a neural network in which hyperparameters will be optimized by a PGA in a Genetically Reinforced Learning scheme. Knowing how the PGA may interact with an environment and work on even very complicated functions this optimization task will be an easy step for those who have already seen the neural network. By the end of the workshop participants will have a solid understanding of how to implement and apply parallel genetic algorithms in Python with practical insights into their strengths and limitations. They will be equipped with the knowledge to extend these techniques to other domains fostering innovation in computational problem-solving.

Real-Time Anomaly Detection and Alerting in Financial Markets Using Stream Processing

Room D4

Tomáš Neubauer, Quix
Tun Shwe, Quix

In the world of financial markets the ability to detect and act on anomalies in real-time is crucial. This workshop will explore how to build a stream processing system that not only detects rapid changes in stock prices but also calculates key stock market indicators like the Relative Strength Index (RSI) Moving Average Convergence Divergence (MACD) or Bollinger Bands in real-time. Attendees will learn how to calculate these indicators in real-time to identify potential buy or sell signals and trigger instant alerts such as Slack messages to notify users of significant market movements or even directly call API to buy/sell instruments. At the end we will discuss and later build a stream processing pipeline in the IDE using the ML model.

A practical guide to LLM-based AI agents

Room D6

Philipp Wendland, Deloitte Consulting

This hands-on workshop is designed to provide participants with an in-depth practical understanding of how to leverage Large Language Models (LLMs) to create intelligent AI agents. As the rise of generative AI continues to transform industries it’s becoming increasingly important for both AI professionals and business leaders to understand the capabilities and implementation strategies for LLM-based systems. The Deloitte AI Institute focusses on brining AI expertise to clients across all industries ranging from innovation over strategy to capability building and scaling. Phillips’s strong technical background in physics / computer science enables him to bridge the gap between business and technology. 1: Introduction to LLM-based agents - Overview of the concept and architecture of AI agents - Introduction to popular frameworks to expedite agent development 2: Hands-on implementation - Guide participants through building a simple LLM-based AI agent using a given framework - Allow for customisation to demonstrate the flexibility and effectiveness of AI agents 3: Industry-Application - Outlook on newest developments of AI agents across various industries - Outlook on the potential of generative Ai and AI agents in particular across industries By the end of the workshop participants will have a solid understanding of the concepts behind LLM-based AI agents and hands-on experience in building these systems themselves. Led by an experienced facilitator from Deloitte’s AI Institute this workshop promises to provide valuable skills and knowledge that participants can leverage in their careers. This workshop is ideal for AI practitioners developers and researchers eager to explore the latest advancements in generative AI and apply LLM-driven automation in their respective fields.

Synthetic Data Generation for Embedding Model Fine-Tuning

Room D7

Ondřej Finke, O2/Dataclair
Stefan Josef, O2/Dataclair
Filip Roskovec, O2/Dataclair

Retrieving information from documents in non-English and domain-specific languages presents a challenge for many organizations. While general embedding models are powerful they often fall short when dealing with specialized terminology not encountered in their training data. This workshop offers a practical approach to addressing these issues: using a combination of real and synthetic data to build robust datasets for fine-tuning open embedding models. The workshop consists of two parts. First we provide an overview of embedding models fine-tuning techniques and methods for generating synthetic data tailored to these approaches. In the second part participants will engage in a hands-on session to generate synthetic data for fine-tuning their own models.

Saturday,
Workshops

O2 Universum, Českomoravská 2345/17a, 190 00, Praha (and on-line)

Registration from 9:00

Welcome to ML Prague 2025

Evaluating LLM outputs with humans and LLMs

Ondřej Dušek, MFF Charles University

How well do LLMs perform on text generation tasks, and how can we tell? We present approaches based on annotating individual errors, using human evaluators as well as LLMs. For humans, we introduce our efficient annotation framework and schema. For LLM-based evaluation, we show a metric using an ensemble of open-source LLMs, which includes a reasoning for each annotated error, evaluated on various generation tasks and evaluation aspects (such as accuracy or fluency) and showing high correlation with human annotators. Both approaches allow us to use benchmarks with recent data unseen to LLMs during training, bypassing the data leakage problem that artificially inflates LLMs' performance on commonly used benchmarks.

Data, your worst enemy?

Johan Loeckx, Vrije Universiteit Brussel

Ever more decisions are driven by advanced, nonlinear data analysis, where the validity, correctness, and fairness of the outcomes are often assumed but difficult to guarantee in practice. We increasingly rely on the output of algorithmic systems (broader than just LLMs) without fully understanding how they arrive at their results. Although much attention has been paid to the validity and fairness of individual predictions or models, the broader topic of AI engineering and its impact remains relatively unexplored.

AI system design is primarily performed by humans tasked with ensuring that operationalization aligns with business objectives. Currently, alignment is handled opaquely by Data Scientists and/or quantified through performance and fairness metrics.

However, many mistakes occur during design—such as violating causality, linearity, or independence constraints, or introducing bias through seemingly minor engineering choices—due to ignorance or the inability to manage complexity. These issues are typically undetectable by metrics and difficult for humans to identify because of the complex interactions between decisions.

In this talk, we will demonstrate how existing approaches fall short and explain how we believe a human-centric, knowledge-driven AutoML architecture and methodology can make the design process more scientific and systems more trustworthy.

Our goal is to combine the strengths of humans and machines, making the AutoML process explainable and leveraging domain knowledge in the synthesis of pipelines and features to ensure alignment. The architecture explores several novel ideas: first, the construction of pipelines and deep features is approached in a unified way. Next, synthesis is driven by a shared knowledge system, interactively queried to determine which pipeline operations to use or features to compute. Lastly, the synthesis process makes decisions at runtime using partial solutions and the results of their application on data. This approach enables interactive collaboration between humans and machines.

Lies, Damn Lies and Gen AI

Jon McLoone, Wolfram Research

While there has been much excitement about the potential of large language models (LLMs) to automate tasks that previously required human intelligence or creativity, many early projects have failed because of LLMs’ innate willingness to lie. The presentation explores the nature, cause and consequences of this “hallucination” issue and proposes a solution.

By combining generative AI with more traditional symbolic AI, reliability can be maintained, explainability improved and private knowledge and data injected. The talk will show simple examples of combining language-based thinking with computational thinking to generate solutions that neither could achieve on its own.

An example application of an AI scientific research assistant will be shown that brings together the ideas presented in a most demanding real-world task, where false information is not acceptable.

LUNCH & POSTER SESSION

An introduction to protein structure prediction

Joseph Pareti, Joseph Pareti's AI consulting

Protein structure prediction has become a cornerstone of modern drug discovery, offering crucial insights for drug-target interactions. This report examines three major computational approaches within an integrated drug development workflow. The primary focus is on AlphaFold 3's architecture, illustrated through simplified implementations of its key components - the Pairformer and Diffusion modules. These demonstration programs provide practical insight into how the system transforms amino acid sequences into accurate three-dimensional structures. RosettaFold is analyzed as a complementary approach, highlighting its comparable performance and distinct advantages in protein design applications. The report also evaluates lightweight protein language models as resource-efficient alternatives for organizations with limited computational infrastructure. Through comparative analysis of these approaches, different prediction strategies can be optimally deployed within the drug development pipeline based on specific requirements for accuracy, speed, and available computational resources.

Mammography solved by AI?

Ivan Cimrák, University of Zilina

Breast cancer screening is an indispensable tool in the early detection of one of the most prevalent cancers among women worldwide. However, one persistent challenge lies in accurately diagnosing clusters of microcalcifications (MCs), which are small calcium deposits within the breast tissue. These clusters are notoriously diverse in appearance, making it difficult to distinguish benign from malignant cases with precision. This diagnostic uncertainty often results in unnecessary biopsies, causing undue stress for patients and placing additional burdens on healthcare systems. To address this, we turned to the transformative potential of artificial intelligence (AI), specifically convolutional neural networks (CNNs), to refine the accuracy of breast cancer screening.

Our study investigates two distinct classification approaches using CNNs: a traditional binary method (classifying MCs as either benign or malignant) and a more advanced three-class method (introducing a third category for non-MC cases). Leveraging two robust datasets—the Curated Breast Imaging Subset of the Digital Database for Screening Mammography and the Optimam Database—we identified ResNet-101 as the most effective CNN architecture for this task. To deepen our understanding of the model’s decision-making process, we employed Grad-CAM visualizations, which highlight the regions of mammogram images that most influence the AI’s predictions.

The results revealed a stark contrast between the two approaches. The binary classification model achieved an accuracy of 74.7% and a Matthews correlation coefficient (MCC) of 0.458. While promising, the model displayed notable limitations in interpretability, often relying on surrounding breast tissue rather than focusing directly on the MCs. Additional challenges arose from benign abnormalities, imaging artifacts, breast implants, and excessive black backgrounds in mammogram patches, further complicating accurate predictions.

In contrast, the three-class model brought a leap in performance, achieving an impressive accuracy of 91.7% and an MCC of 0.767. By introducing a third classification category, this model demonstrated an enhanced ability to isolate and evaluate microcalcifications with greater precision. However, this advancement came with a new challenge: vascular calcifications, which were not adequately represented in the training datasets, were occasionally misclassified. These findings underscore the importance of robust, comprehensive datasets to ensure accurate and reliable AI models.

Our research highlights the immense potential of AI to revolutionize breast cancer screening by improving diagnostic accuracy, reducing unnecessary procedures, and enhancing the overall efficiency of healthcare delivery. The significant improvement achieved with the three-class classification approach offers a glimpse into a future where AI can act as a trusted partner for radiologists. However, our findings also emphasize the critical need for continuous refinement of these systems. Addressing the misclassification of vascular calcifications and expanding datasets to encompass a broader range of imaging scenarios will be pivotal in unlocking the full potential of this technology. In this era of rapid technological advancement, our study takes a vital step toward combining cutting-edge AI with human expertise to create a more precise and patient-centered approach to breast cancer detection.

End-to-end Stroke imaging analysis, using reservoir computing-based effective connectivity, and interpretable AI

Alessandro Crimi, AGH University of Krakow

We present an End-to-end AI framework for directed graphs including explainable AI.

This is a machine learning pipeline combining reservoir computing and directed graph analysis to model brain connectivity in stroke patients using MRI data. Effective connectivity is derived via reservoir computing, enabling the creation of directed graph representations. These graphs are classified using a directed graph convolutional networ. Explainable AI tools provide insights into disrupted brain networks, elucidating biomarkers for stroke classification and enhancing clinical interpretability. This approach highlights the potential of machine learning to improve patient stratification in stroke and other brain diseases. The technical innovations are related to reservoir computing networks, directed graph analysis, and explainable AI of effective brain connectivity.

COFFEE BREAK

Attentive interpretable networks for scalable content recommendation in mobile games

Martin Dlask, King

Candy Crush Saga, a popular mobile game with millions of monthly players, leverages AI to ensure gameplay remains engaging, relevant, and fair. The talk consists of two parts. The first part focuses on introducing a novel approach for content recommendation using attentive networks, which was published in RecSys in 2024. We describe the architecture, experimental results, and a scale-adaptive algorithm for fair and relevant recommendations of offers and rewards for completing in-game quests. The second part outlines the implementation of a scalable prediction system for millions of players. We introduce a serving architecture for mobile games, including a typical client-server model. Alongside the talk, we present practical advice on ensuring a robust implementation of recommender systems: detecting degenerate feedback loops, preventing loss of relevance over time, overcoming cold-start problems, and removing bias.

Evolution of Recommendation System: from ANN to Ensemble of Scorers

Raid Arfua, GR8 Tech

This talk draws from my recent article, which chronicles the evolution of a sports event recommendation system from its early neural network based approach — implemented well before Deep Learning gained mainstream popularity — into a more effective ensemble of simpler scoring methods. These methods range from traditional statistical formulas to sophisticated similarity algorithms leveraging sparse matrix structures.

In this session, we will delve into the technical details of working with sparse matrices, as well as the development and application of the “Hyperbolic Score” evaluation metric, examining why it can be particularly effective under certain conditions. We will also discuss the importance of building a robust Data Platform as a foundation for rapid experimentation, ultimately enabling more efficient development and iteration of recommendation engines.

Beyond the established techniques, I will introduce several thought-provoking ideas for integrating large language models (LLMs) into recommender systems. Additionally, we will explore strategies for balancing model complexity with interpretability, ensuring that both accuracy and transparency remain central concerns in model design.

My hope is that these insights and practical considerations will provide the ML and Data Science community with valuable perspectives on building robust, scalable, and explainable Recommendation Systems.

Estimating online behavior of ad hoc cohorts using context-dependent weighing of panel participants

Ariel Azia, Similarweb

As much of human activity has moved to the internet in the last few decades, businesses must take into account behavior of individuals and groups online; trends of website or mobile application usage, volume of terms used in search engines, and the popularity pages of specific products and services can all contribute to an understanding of online activity patterns, allowing companies to make informed strategic decisions. Similarweb is one such company to provide these digital insights, termed digital business intelligence, for a myriad of customers and use cases. In the process of generating these metrics, Similarweb combines a small set of actual data with more extensive information from a selected panel of online users. 

A particularly difficult task for Similarweb is the estimation of ad hoc cohorts of users, as clients require the freedom to query and compare their relevant online behavior metric for their specific use case. As it is almost impossible to generate a priori estimations of every possible cohort, it is substantially useful to measure the interaction between each specific user within the cohort and the cohort itself and assign it an individual weight. This weighing problem is similar in essence to problems in recommendation, whose solutions are well established,  but is further complicated by sampling biases within the available actual data, biases within the panel users, the sparsity of interactions between users and cohorts, and the multitude of possible cohorts. We briefly present an earlier approach we have adopted i.e. assigning each panel user a singular weight and applying simple rescaling within a given cohort. This singular weight is useful under strong constraints, but yields poor results on ad hoc cohorts as it cannot account for the non linear nature of interaction between users and cohorts of which they are a part.

We introduce a new multistep approach to create the context dependent weighing. First, we obtain a representative embedding of both common types of cohorts (websites, search terms, product views etc.) and a respective embedding representing users using the same encoder. Second, we train a recommendation-like neural network that learns the non linear interactions between users and their cohorts. This new approach allows us to obtain both an overall sum representation as well as the inner weight distribution for an ad hoc cohort. We demonstrate the usefulness of this new approach on several examples of ad hoc cohorts. It is interesting to note that as an additional byproduct of this training process, we can extract a useful intermediate from the network that embeds both users and cohorts under the constraint of actual data and panel biases.

COFFEE BREAK

The Evolution of Virtual Buddy: From Concept to Deployment

Alexandr Vendl, Dataclair

Intelligent Manufacturing Assistant Bot

Alexander Jesser, University Heilbronn

In a joint research project of the Institute for Intelligent Cyber-Physical Systems ICPS at Heilbronn University and the  industrial partners, SABO Mobile IT GmbH and Grossebacher Systeme AG, research is being conducted into a middleware for voice assistants for global auditory human-technology interaction in the field of industrial plants.
Interactive voice bots are becoming increasingly common in consumer devices and have great potential to support operators and service personnel in operating complex machines. Due to their lack of user-friendliness, flexibility, security and integration into existing industrial solutions, conventional command controls have so far only met with limited acceptance in industrial environments. Chatbots as well as voicebots both involve so-called Natural Language Understanding (NLU) and Natural Language Processing (NLP). However, with voicebots, much more emphasis must be placed on conversation design in order to convey the conversation content in a compact yet understandable way. In addition, a robust method for recognizing voice commands is required. Thanks to their comparatively lower complexity chatbots are already widely used in a variety of applications. Chatbots are mainly used for processes that are easy to automate and always follow a similar pattern. Voicebots, on the other hand, are not or hardly ever used in industrial production environments. The ambient noise level and process reliability are major hurdles. Furthermore, natural communication is not possible with current voicebots. It can be summarized that voicebots are mainly used in acoustically quiet and less critical. 
The aim of the joint research is to realize an assistant for machines that informs the operator, takes over his instructions, checks the exchanged information and thus ensures that a machine works according to the operator's wishes and within its technical capabilities. The deep integration of the intelligent manufacturing assistant bot (IMAB) into the machine software and the understanding of natural language communication is the key to controlling the machine, in contrast to a limited set of commands that humans, not machines, have to learn. The scientific innovation lies in an intelligent language assistant capable of natural language dialogs beyond the level of regular chat bots. A new intermediate layer between a language chat bot-oriented user interface and machine control enables hands-free language communication in combination with classic user interfaces.
The research path will be demonstrated in this talk. Aspects of the resource-related hardware architecture required in an industrial environment and the intelligent software architectures will be presented. In particular, intelligent noise cancellation in the speech signal under a wide range of industrial ambient noise is of crucial importance. Furthermore, the integrated speech-to-text (STT) and text-to-speech (TTS) processes are presented.
In addition, the implementation of an intelligent speaker recognition based on neural networks with third-party speaker detection and an intelligent user authentication based on voice input is demonstrated.
The presentation will cover the latest research results of the process steps within the project requirements. In particular, the scientifically innovative core issues of the research, which are essentially based on machine learning methods, will be demonstrated.

PARTY

Sunday,
Conference day 1

O2 Universum, Českomoravská 2345/17a, 190 00, Praha (and on-line)

Doors open at 08:30

Solving adversarial attacks in computer vision

Stanislav Fort, Google DeepMind

Training AI Models for Crime Scene Fingerprint Recognition

Jakub Sochor, Innovatrics

This talk delves into critical aspects of latent fingerprint recognition in crime scene investigations, focusing on fingerprint analysis and system evaluation. We will discuss the challenge of training minutiae detectors without ground truth annotations and introduce innovative approaches using synthetically generated fingerprint data. By leveraging AI and synthetic data, we will show how these advancements can significantly enhance the accuracy and efficiency of fingerprint recognition in forensic science.

Understanding the neural networks through rule extraction

Tomas Pevny, Czech Technical University

Neural networks are ubiquitous yet they remain opaque for most of its users, who has very little understanding of how they store the knowledge and how the information propagates through. In this talk, I would like to share our findings from our quest to understand these phenomena. Specifically I will show the decision rules realized by neural networks and why it might be difficult to understand them without the knowledge of the data distribution. This will give us intuition why neural networks are robust yet why adversarial samples are so easy to create. Finally, we will use these tools to understand, how the decision rules compose during inference.

COFFEE BREAK

Distributed Collaborative AI

Hava Siegelmann, University of Massachusetts Amherst

Advances and Challenges in Topic Modeling of Text Documents

Martin Neznal, Productboard

In this talk, we will explore the field of topic modeling for text documents, focusing on its challenges and practical applications. I will highlight various methods for clustering text documents, enhancing clustering quality, validating results, and integrating solutions into users' daily workflows. Our presentation will share insights and lessons learned from building topic modeling inside a product that serves real customers, emphasizing the challenges of creating scalable systems that deliver real value.

I will emphasize the importance of preprocessing input documents to improve clustering quality. This includes extracting relevant elements from the text, such as entities, key phrases, or summaries. Subsequently, I will demonstrate and compare methods for clustering these text representations, such as hierarchical clustering or directly using LLMs to cluster a large set of documents of text.
In real-world scenarios, new text documents are created daily, and new clusters emerge over time. We will explore techniques to detect new clusters as they appear while maintaining the integrity of existing clusters.

Assessing the quality of discovered topics is a key challenge in topic modeling. I will provide an overview of validation techniques, ranging from traditional machine learning metrics to methods that use LLM as a judge. Furthermore, we will discuss the importance of human-in-the-loop validation processes in ensuring the relevance and accuracy of the topics.

Finally, I will share insights on improving the usability and user experience of topic modeling, including effective naming and description of clusters, and we will discuss whether users are willing to provide feedback on AI models and how this feedback can be used to refine and enhance the existing solutions.

Towards Real-World Fact-Checking with Large Language Models

Iryna Gurevych, Technical University of Darmstadt

Misinformation poses a growing threat to our society. It has a severe impact on public health by promoting fake cures or vaccine hesitancy, and it is used as a weapon during military conflicts to spread fear and distrust. Current research on natural language processing (NLP) for fact-checking focuses on identifying evidence and predicting the veracity of a claim. People's beliefs, however, often do not depend on the claim and rational reasoning but on credible content that makes the claim seem more reliable, such as scientific publications or visual content that was manipulated or stems from unrelated contexts. To combat misinformation, we need to show (1) "Why was the claim believed to be true?", (2) "Why is the claim false?", (3) "Why is the alternative explanation correct?". In this talk, I will zoom in on two critical aspects of such misinformation supported by credible though misleading content. Firstly, I will present our efforts to dismantle misleading narratives based on fallacious interpretations of scientific publications. Secondly, I will show how we can use multimodal large language models to (1) detect misinformation based on visual content and (2) provide strong alternative explanations for the visual content.

LUNCH & POSTER SESSION

Towards Production-Ready Czech LLMs with Continuous Pretraining

Ondřej Filip, Seznam.cz

Seznam has been developing custom large language models (LLMs) with a focus on the Czech language and the wide range of Internet services we provide. At the dawn of open-weight models, these early systems lacked several critical capabilities, such as generating fluent and natural Czech text efficiently, understanding nuanced cultural contexts, and managing long-form content. This talk will explore how we addressed these limitations during that formative period and share insights from our journey toward production-ready Czech LLMs.

How to feed your LLMs with data from the web

Jan Čurn, Apify

All major generative AI and Large Language Models (LLMs) have been trained using data scraped from the web.

Additionally, LLM applications often extract web data to provide up-to-date context for answers using Retrieval Augmented Generation (RAG).

However, reliably collecting online data at scale is challenging due to issues like blocking, dynamic content rendering, and the sheer volume of data.

In this talk, we will explain how you can establish an efficient web data extraction pipeline, clean the HTML to circumvent the “garbage in, garbage out” problem, and present examples of successful applications.

Fitting LLMs into a single GPU: Making neural networks smaller

Vladimir Macko, GrizzlyTech, former Google AI

As neural networks continue to grow in size and complexity, the demand for efficient models has never been greater. Neural network pruning and quantization have emerged as two of the most promising techniques for reducing model size and improving computational efficiency. But how do these techniques translate from theoretical research to real-world applications?

In this talk, we will explore the state of the art in neural network pruning and quantization, presenting key findings from academia alongside lessons learned from industry implementations.

Using examples from real-world projects, we will discuss practical approaches to algorithm selection, toolchain optimization, and model evaluation. Whether you're a machine learning researcher or a practitioner, this session will equip you with actionable strategies to make neural networks faster, smaller, and more efficient without compromising performance.

Join us to bridge the gap between cutting-edge research and applied machine learning, and discover how to make the most of these transformative techniques in your work.

COFFEE BREAK

PANEL DISCUSSION

CLOSING REMARKS

Have a great time Prague, the city that never sleeps

You can feel centuries of history at every corner in this unique capital. We'll invite you to get a taste of our best pivo (that’s beer in Czech) and then bring you back to the present day at our networking event.

Venue ML Prague 2025 will run hybrid, in person and online!

The main conference as well as the workshops will be held at O2 Universum.

We will also livestream the talks for all those participants who prefer to attend the conference online. Our platform will allow interaction with speakers and other participants too. Workshops require intensive interaction and won't be streamed.

Conference building

O2 Universum
Českomoravská 2345/17a, 190 00, Praha 9

Workshops

O2 Universum
Českomoravská 2345/17a, 190 00, Praha 9

Now or never Registration

Early Bird

Sold Out

  • Conference days € 270
  • Only workshops € 200
  • Conference + workshops € 440

Standard

Late

Until sold out

  • Conference days € 320
  • Only workshops € 260
  • Conference + workshops € 520

What You Get

  • Practical and advanced level talks led by top experts.
  • Networking and drinks with speakers and people from all around the world.
  • Delicious food and snacks throughout the conference.

They’re among us We are in The ML Revolution age

Machines can learn. Incredibly fast. Faster than you. They are getting smarter and smarter every single day, changing the world we’re living in, our business and our life. The artificial intelligence revolution is here. Come, learn and make this threat your biggest advantage.

Our Attendees What they say about ML Prague

Thank you to Our Partners

Strategic Partners

Platinum Partners

Gold Partners

Communities and Further support

Would you like to present your brand to 1000+ Machine Learning enthusiasts? Send us an email at info@mlprague.com to find out how to become a ML Prague 2025 partner. Our basic partnership offer can be found here.

Become a partner

Happy to help Contact

If you have any questions about Machine Learning Prague, please e-mail us at
info@mlprague.com

Organizers

Jiří Materna
Scientific program & Co-Founder
jiri@mlprague.com

Teresa Pulda
Event production
teresa@mlprague.com

Gonzalo V. Fernández
Marketing and social media
gonzalo@mlprague.com

Jona Azizaj
Partnerships
jona@mlprague.com

Ivana Javná
Speaker support
ivana@mlprague.com

Barbora Toman Hanousková
Communication
barbora@mlprague.com

Jan Romportl
Moderator